32 research outputs found

    Laying the Foundation for In-car Alcohol Detection by Speech

    Get PDF
    The fact that an increasing number of functions in the automobile are and will be controlled by speech of the driver rises the question whether this speech input may be used to detect a possible alcoholic intoxication of the driver. For that matter a large part of the new Alcohol Language Corpus (ALC) edited by the Bavarian Archive of Speech Signals (BAS) will be used for a broad statistical investigation of possible feature candidates for classification. In this contribution we present the motivation and the design of the ALC corpus as well as first results from fundamental frequency and rhythm analysis. Our analysis by comparing sober and alcoholized speech of the same individuals suggests that there are in fact promising features that can automatically be derived from the speech signal during the speech recognition process and will indicate intoxication for most speakers

    Mobile learning — trends and practices

    Get PDF
    Mobile learning has become one of the more influential aspects in the field of educational technology given the ubiquity of modern mobile devices and proliferation of educational applications or ‘apps’ for mobile devices. Within this special issue, there are a range of studies and reviews which cover a breadth of current topics in the field, namely user motivations for using mobile learning, issues in evaluation and domain-specific considerations (e.g., use within language learning or audio-based applications). Together these studies represent the synthesis of a range of methods, approaches and applications that highlight benefits and areas of future growth of mobile technologies

    The Role of Subjective Quality Judgements in User Preferences for Mobile Learning Apps

    Get PDF
    This study investigated whether subjective quality judgements on sound and picture quality across three devices (iPhone, iPad, and iPad mini) affected user preferences for learning applications. We tested 20 native Finnish-speaking users trialing generic audio clips, video clips, and two kinds of learning apps that were heavily reliant on sound. It was found that there was a main effect of the device on perceived sound quality, replicating earlier findings. However, these judgements did not impact on the users' preferences for different devices nor on their preferences for different applications. The results are interpreted as indicating that perceived quality and affordances are less important for users in these contexts than other considerations (e.g., convenience, mobility, etc.).Peer reviewe

    Mobile Learning

    Get PDF
    Mobile learning has become one of the more influential aspects of the field of educational technology, given the ubiquity of modern mobile devices and proliferation of educational applications or ‘apps’. Within this volume, there are a range of studies and reviews which cover a breadth of current topics in the field, namely user motivations for using mobile learning, issues in evaluation, as well as domain-specific considerations (e.g., use within language learning or audio-based applications). Together, these studies represent the synthesis of a range of methods, approaches, and applications that highlight benefits and areas of future growth of mobile technologies and how they can be useful and most effective in education

    Phonetic and orthographic cues are weighted in speech sound perception by second language speakers: evidence from Greek speakers of English.

    Get PDF
    Second language (L2) learning can involve processing of speech-sound contrasts that have multiple phonetic cues (e.g. Iverson et al., 2003). This can be particularly difficult for foreign-language learners especially if the cues are weighted differently in the foreign and native languages (e.g., Giannakopoulou et al., 2011, 2013). The orthographic representation of words is suggested to also interfere with speech sound perception in way of presenting additional cues for the second language learner. Greek child and adult speakers of English were studied to determine on what basis they are making perceptual identification between English vowels with the use of pictures as visual stimuli. Performance was impaired for Greek speakers across all tasks but worst for Greek speakers for the picture stimuli task. Findings suggest a 'link' between orthography and perceptual identification serving as an additional cue for L2 speakers

    How does foreigner-directed speech differ from other forms of listener-directed clear speaking styles?

    Get PDF
    Forty talkers participated in problem-solving tasks with another talker in conditions differing in communication difficulty for the interlocutor. A linguistic barrier condition (L2 interlocutor) was compared to acoustic barrier conditions (native interlocutors hearing vocoded or noisy speech). Talkers made acoustic-phonetic enhancements in all barrier conditions compared to the no-barrier condition, but talkers reduced their articulation rate less and showed less increase in vowel hyperarticulation in foreigner-directed speech than in the acoustic barrier condition, even though communicative difficulty was greater in the L2 condition. Foreigner-directed speech was also perceived as less clear. This suggests that acoustic enhancements in clear speech are not simply a function of the level of communication difficulty

    The effect of HVP training in vowel perception on bilingual speech production

    Get PDF
    This is an accepted manuscript of an article published by Conscientia Beam in Research in English Language Teaching, available online at: http://www.conscientiabeam.com/archive/179/06-2021/1 The accepted version of the publication may differ from the final published version.Prior investigations (Giannakopoulou et al., 2013) have indicated high variability phonetic training intervention can help L2 English adult learners change the perception of vowels such that they shift their attention to primary cues (spectral features) rather than secondary cues (e.g. duration) to correctly identify vowels in L2. This experiment explores if high-variability training impacts on L2 adult learners’ production of L2 speech. Production samples from a prior experiment were used to conduct ratings of accuracy (Giannakopoulou, 2012). In the current experiment, the production samples were transcribed and rated for accuracy by twenty native English listeners. The intelligibility levels of L2 learners’ speech samples as indexed by higher accuracy in transcription were observed as having been rated higher following training than prior to training. The implications of the results are considered with regard to theories on the connection between speech production and perception, and Flege’s (1995) Speech Learning Model

    The effect of hyperarticulation on speech comprehension under adverse listening conditions

    Get PDF
    © 2021 The Authors. Published by Springer. This is an open access article available under a Creative Commons licence. The published version can be accessed at the following link on the publisher’s website: https://doi.org/10.1007/s00426-021-01595-2Comprehension assesses a listener’s ability to construe the meaning of an acoustic signal in order to be able to answer questions about its contents, while intelligibility indicates the extent to which a listener can precisely retrieve the acoustic signal. Previous comprehension studies asking listeners for sentence-level information or narrative-level information used native listeners as participants. This is the first study to look at whether clear speech properties (e.g. expanded vowel space) produce a clear speech benefit at the word level for L2 learners for speech produced in naturalistic settings. This study explored whether hyperarticulated speech was more comprehensible than non-hyperarticulated speech for both L1 British English speakers and early and late L2 British English learners in quiet and in noise. Sixteen British English listeners, 16 native Mandarin Chinese listeners as early learners of L2 and 16 native Mandarin Chinese listeners as late learners of L2 rated hyperarticulated samples versus non-hyperarticulated samples in form of words for comprehension under four listening conditions of varying white noise level (quiet or SNR levels of +16dB, +12dB or +8dB) (3x2x4 mixed design). Mean ratings showed all three groups found hyperarticulated speech samples easier to understand than non-hyperarticulated speech at all listening conditions. Results are discussed in terms of other findings (Uther et al., 2012) that suggest that hyperarticulation may generally improve speech processing for all language groups.The studies were funded by an Isambard Scholarship from Brunel University

    User perceptions of sound quality: implications for the design and use of audio-based mobile applications

    Get PDF
    This study sought to investigate the effect that contextual cues (in particular, device type and content type) have on the perception of sound quality. A sample of 49 participants were tested on different mobile devices sizes (small – iPhone, medium – iPad Mini, and large – iPad) which had identical sound output characteristics within in different usage contexts (generic content vs. musical training app contexts). Results showed that the users’ perception of generic sound types was affected by device type, with iPhones appearing to have better sound quality compared to larger devices. On the other hand, within application contexts, the application type seemed to affect user perceptions more, with the rhythm training application rating poorer on sound quality, picture quality, and likelihood of future use as compared to the pitch training application (although this may be due to the perceived increased difficulty). Together, these findings demonstrate the influence of device and content cues (when actual physical qualities are controlled) on user sound perception. Interestingly, differences in perceived sound quality was not accompanied by an overriding preference for that device as compared to other devices. Instead, considerations such as ease of use seemed to drive considerations for uptake of applications

    User Experiences from L2 Children Using a Speech Learning Application : Implications for Developing Speech Training Applications for Children

    Get PDF
    We investigated user experiences from 117 Finnish children aged between 8 and 12 years in a trial of an English language learning programme that used automatic speech recognition (ASR). We used measures that encompassed both affective reactions and questions tapping into the children' sense of pedagogical utility. We also tested their perception of sound quality and compared reactions of game and nongame-based versions of the application. Results showed that children expressed higher affective ratings for the game compared to nongame version of the application. Children also expressed a preference to play with a friend compared to playing alone or playing within a group. They found that assessment of their speech is useful although they did not necessarily enjoy hearing their own voices. The results are discussed in terms of the implications for user interface (UI) design in speech learning applications for children.Peer reviewe
    corecore